Efficient Crowdsourcing of Unknown Experts using Multi-Armed Bandits
نویسندگان
چکیده
We address the expert crowdsourcing problem, in which an employer wishes to assign tasks to a set of available workers with heterogeneous working costs. Critically, as workers produce results of varying quality, the utility of each assigned task is unknown and can vary both between workers and individual tasks. Furthermore, in realistic settings, workers are likely to have limits on the number of tasks they can perform and the employer will have a fixed budget to spend on hiring workers. Given these constraints, the objective of the employer is to assign tasks to workers in order to maximise the overall utility achieved. To achieve this, we introduce a novel multi–armed bandit (MAB) model, the bounded MAB, that naturally captures the problem of expert crowdsourcing. We also propose an algorithm to solve it efficiently, called bounded ε–first, which uses the first εB of its total budget B to derive estimates of the workers’ quality characteristics (exploration), while the remaining (1 − ε) B is used to maximise the total utility based on those estimates (exploitation). We show that using this technique allows us to derive an O ( B 2 3 ) upper bound on our algorithm’s performance regret (i.e. the expected difference in utility between the optimal and our algorithm). In addition, we demonstrate that our algorithm outperforms existing crowdsourcing methods by up to 155% in experiments based on real– world data from a prominent crowdsourcing site, while achieving up to 75% of a hypothetical optimal with full information.
منابع مشابه
Efficient crowdsourcing of unknown experts using bounded multi-armed bandits
Increasingly, organisations flexibly outsource work on a temporary basis to a global audience of workers. This so-called crowdsourcing has been applied successfully to a range of tasks, from translating text and annotating images, to collecting information during crisis situations and hiring skilled workers to build complex software. While traditionally these tasks have been small and could be ...
متن کاملFrom Bandits to Experts: A Tale of Domination and Independence
We consider the partial observability model for multi-armed bandits, introduced by Mannor and Shamir [11]. Our main result is a characterization of regret in the directed observability model in terms of the dominating and independence numbers of the observability graph. We also show that in the undirected case, the learner can achieve optimal regret without even accessing the observability grap...
متن کاملFrom Bandits to Experts: A Tale of Domination and Independence
We consider the partial observability model for multi-armed bandits, introducedby Mannor and Shamir [11]. Our main result is a characterization of regret inthe directed observability model in terms of the dominating and independencenumbers of the observability graph. We also show that in the undirected case, thelearner can achieve optimal regret without even accessing the observ...
متن کاملBandits with Knapsacks: Dynamic procurement for crowdsourcing∗
In a basic version of the dynamic procurement problem, the algorithm has a budget B to spend, and is facing n agents (potential sellers) that are arriving sequentially. The algorithm offers a take-it-or-leave-it price to each arriving seller; the sellers value for an item is an independent sample from some fixed (but unknown) distribution. The goal is to maximize the number of items bought. Thi...
متن کاملRotting Bandits
The Multi-Armed Bandits (MAB) framework highlights the tension between acquiring new knowledge (Exploration) and leveraging available knowledge (Exploitation). In the classical MAB problem, a decision maker must choose an arm at each time step, upon which she receives a reward. The decision maker’s objective is to maximize her cumulative expected reward over the time horizon. The MAB problem ha...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2012